S-DFP: shifted dynamic fixed point for quantized deep neural network training

نویسندگان

چکیده

Abstract Recent advances in deep neural networks have achieved higher accuracy with more complex models. Nevertheless, they require much longer training time. To reduce the time, methods using quantized weight, activation, and gradient been proposed. Neural network calculation by integer format improves energy efficiency of hardware for learning Therefore, fixed point However, narrow data representation range degrades accuracy. In this work, we propose a new named shifted dynamic (S-DFP) to prevent degradation training. S-DFP can change adding bias exponent. We evaluated effectiveness on ImageNet task ResNet-34, ResNet-50, ResNet-101 ResNet-152. For example, ResNet-152 is improved from 76.6% conventional 8-bit DFP 77.6% S-DFP.

برای دانلود رایگان متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

منابع مشابه

Deep Convolutional Neural Network Inference with Floating-point Weights and Fixed-point Activations

Deep convolutional neural network (CNN) inference requires significant amount of memory and computation, which limits its deployment on embedded devices. To alleviate these problems to some extent, prior research utilize low precision fixed-point numbers to represent the CNN weights and activations. However, the minimum required data precision of fixed-point weights varies across different netw...

متن کامل

Overcoming Challenges in Fixed Point Training of Deep Convolutional Networks

It is known that training deep neural networks, in particular, deep convolutional networks, with aggressively reduced numerical precision is challenging. The stochastic gradient descent algorithm becomes unstable in the presence of noisy gradient updates resulting from arithmetic with limited numeric precision. One of the wellaccepted solutions facilitating the training of low precision fixed p...

متن کامل

Deep neural network training emphasizing central frames

It is common practice to concatenate several consecutive frames of acoustic features as input of a Deep Neural Network (DNN) for speech recognition. A DNN is trained to map the concatenated frames as a whole to the HMM state corresponding to the center frame while the side frames close to both ends of the concatenated frames and the remaining central frames are treated as equally important. Tho...

متن کامل

$S$-metric and fixed point theorem

In this paper, we prove a general fixed point theorem in $textrm{S}$-metric spaces for maps satisfying an implicit relation on complete metric spaces. As applications, we get many analogues of fixed point theorems in metric spaces for $textrm{S}$-metric spaces.

متن کامل

Mixed Low-precision Deep Learning Inference using Dynamic Fixed Point

We propose a cluster-based quantization method to convert pre-trained full precision weights into ternary weights with minimal impact on the accuracy. In addition we also constrain the activations to 8-bits thus enabling sub 8-bit full integer inference pipeline. Our method uses smaller clusters of N filters with a common scaling factor to minimize the quantization loss, while also maximizing t...

متن کامل

ذخیره در منابع من


  با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید

ژورنال

عنوان ژورنال: Neural Computing and Applications

سال: 2021

ISSN: ['0941-0643', '1433-3058']

DOI: https://doi.org/10.1007/s00521-021-06821-x